Sensor-Based Navigation Using Hierarchical Reinforcement Learning
نویسندگان
چکیده
Robotic systems are nowadays capable of solving complex navigation tasks. However, their capabilities limited to the knowledge designer and consequently lack generalizability initially unconsidered situations. This makes deep reinforcement learning (DRL) especially interesting, as these algorithms promise a self-learning system only relying on feedback from environment. In this paper, we consider problem lidar-based robot in continuous action space using DRL without providing any goal-oriented or global information. By solely local sensor data solve tasks, design an agent that assigns its own waypoints based intrinsic motivation. Our is able learn goal-directed behavior even when facing sparse feedback, i.e., delayed rewards reaching target. To address challenge complexity space, deploy hierarchical structure which exploration distributed across multiple layers. Within structure, our self-assigns internal goals learns extract reasonable reach desired target position data. experiments, demonstrate two environments show seriously improves performance terms success rate weighted by path length comparison flat structure. Furthermore, provide real-robot experiment illustrate trained can be easily transferred real-world scenario.
منابع مشابه
Hierarchical Reinforcement Learning for Robot Navigation
For complex tasks, such as manipulation and robot navigation, reinforcement learning (RL) is well-known to be difficult due to the curse of dimensionality. To overcome this complexity and making RL feasible, hierarchical RL (HRL) has been suggested. The basic idea of HRL is to divide the original task into elementary subtasks, which can be learned using RL. In this paper, we propose a HRL archi...
متن کاملRobot Navigation in Partially Observable Domains using Hierarchical Memory-Based Reinforcement Learning
In this paper, we attempt to find a solution to the problem of robot navigation in a domain with partial observability. The domain is a grid-world with intersecting corridors, where the agent learns an optimal policy for navigation by making use of a hierarchical memory-based learning algorithm. We define a hierarchy of levels over which the agent abstracts the learning process, as well as it...
متن کاملTree Based Hierarchical Reinforcement Learning
In this thesis we investigate methods for speeding up automatic control algorithms. Specifically, we provide new abstraction techniques for Reinforcement Learning and Semi-Markov Decision Processes (SMDPs). We introduce the use of policies as temporally abstract actions. This is different from previous definitions of temporally abstract actions as we do not have termination criteria. We provide...
متن کاملHierarchical Memory-Based Reinforcement Learning
A key challenge for reinforcement learning is scaling up to large partially observable domains. In this paper, we show how a hierarchy of behaviors can be used to create and select among variable length short-term memories appropriate for a task. At higher levels in the hierarchy, the agent abstracts over lower-level details and looks back over a variable number of high-level decisions in time....
متن کاملHierarchical Explanation-Based Reinforcement Learning
Explanation-Based Reinforcement Learning (EBRL) was introduced by Dietterich and Flann as a way of combining the ability of Reinforcement Learning (RL) to learn optimal plans with the generalization ability of Explanation-Based Learning (EBL) (Di-etterich & Flann, 1995). We extend this work to domains where the agent must order and achieve a sequence of subgoals in an optimal fashion. Hierarchi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture notes in networks and systems
سال: 2023
ISSN: ['2367-3370', '2367-3389']
DOI: https://doi.org/10.1007/978-3-031-22216-0_37